Supervised Multivariate Learning with Simultaneous Feature Auto-Grouping and Dimension Reduction
نویسندگان
چکیده
Abstract Modern high-dimensional methods often adopt the ‘bet on sparsity’ principle, while in supervised multivariate learning statisticians may face ‘dense’ problems with a large number of nonzero coefficients. This paper proposes novel clustered reduced-rank (CRL) framework that imposes two joint matrix regularizations to automatically group features constructing predictive factors. CRL is more interpretable than low-rank modelling and relaxes stringent sparsity assumption variable selection. In this paper, new information-theoretical limits are presented reveal intrinsic cost seeking for clusters, as well blessing from dimensionality learning. Moreover, an efficient optimization algorithm developed, which performs subspace clustering guaranteed convergence. The obtained fixed-point estimators, although not necessarily globally optimal, enjoy desired statistical accuracy beyond standard likelihood setup under some regularity conditions. kind information criterion, its scale-free form, proposed cluster rank selection, has rigorous theoretical support without assuming infinite sample size. Extensive simulations real-data experiments demonstrate interpretability method.
منابع مشابه
Supervised dimension reduction mappings
Abstract. We propose a general principle to extend dimension reduction tools to explicit dimension reduction mappings and we show that this can serve as an interface to incorporate prior knowledge in the form of class labels. We explicitly demonstrate this technique by combining locally linear mappings which result from matrix learning vector quantization schemes with the t-distributed stochast...
متن کاملIntrusion Detection using Supervised Learning with Feature Set Reduction
Intrusion detection systems intend to recognize attacks with a low false positive rate and high detection rate. Many feature selection methods introduced to eliminate redundant and irrelevant features, because raw features may abbreviate accuracy or robustness of classification. In this paper we are proposing the information gain technique for the selection of the features. A feature with the h...
متن کاملSupervised dimension reduction with topic models
We consider supervised dimension reduction (SDR) for problems with discrete variables. Existing methods are computationally expensive, and often do not take the local structure of data into consideration when searching for a low-dimensional space. In this paper, we propose a novel framework for SDR which is (1) general and flexible so that it can be easily adapted to various unsupervised topic ...
متن کاملGradient-based kernel dimension reduction for supervised learning
This paper proposes a novel kernel approach to linear dimension reduction for supervised learning. The purpose of the dimension reduction is to find directions in the input space to explain the output as effectively as possible. The proposed method uses an estimator for the gradient of regression function, based on the covariance operators on reproducing kernel Hilbert spaces. In comparison wit...
متن کاملLinear Contour Learning: A Method for Supervised Dimension Reduction
We propose a novel approach to sufficient di mension reduction in regression, based on es timating contour directions of negligible vari ation for the response surface. These di rections span the orthogonal complement of the minimal space relevant for the regression, and can be extracted according to a mea sure of the variation in the response, lead ing to General Contour Regression (GCR)...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of The Royal Statistical Society Series B-statistical Methodology
سال: 2022
ISSN: ['1467-9868', '1369-7412']
DOI: https://doi.org/10.1111/rssb.12492